Efficient Approximation for Large-Scale Kernel Clustering Analysis
نویسندگان
چکیده
Kernel k-means is useful for performing clustering on nonlinearly separable data. The kernel k-means is hard to scale to large data due to the quadratic complexity. In this paper, we propose an approach which utilizes the low-dimensional feature approximation of the Gaussian kernel function to capitalize a fast linear k-means solver to perform the nonlinear kernel k-means. This approach takes advantage of the efficiency of the linear solver and the nonlinear partitioning ability of the kernel clustering. The experimental results show that the proposed approach is much more efficient than a normal kernel kmeans solver and achieves similar clustering performance.
منابع مشابه
Kernel-Based Clustering of Big Data
There has been a rapid increase in the volume of digital data over the recent years. Analysis of this data, popularly known as big data, necessitates highly scalable data analysis techniques. Clustering is an exploratory data analysis tool used to discover the underlying groups and structures in the data. Stateof-the-art scalable clustering algorithms assume “linear separability” of the cluster...
متن کاملMemory Efficient Kernel Approximation
Scaling kernel machines to massive data sets is a major challenge due to storage and computation issues in handling large kernel matrices, that are usually dense. Recently, many papers have suggested tackling this problem by using a low-rank approximation of the kernel matrix. In this paper, we first make the observation that the structure of shift-invariant kernels changes from low-rank to blo...
متن کاملApproximate kernel competitive learning
Kernel competitive learning has been successfully used to achieve robust clustering. However, kernel competitive learning (KCL) is not scalable for large scale data processing, because (1) it has to calculate and store the full kernel matrix that is too large to be calculated and kept in the memory and (2) it cannot be computed in parallel. In this paper we develop a framework of approximate ke...
متن کاملScalable Kernel K-Means Clustering with Nystrom Approximation: Relative-Error Bounds
Kernel k-means clustering can correctly identify and extract a far more varied collection of cluster structures than the linear k-means clustering algorithm. However, kernel kmeans clustering is computationally expensive when the non-linear feature map is highdimensional and there are many input points. Kernel approximation, e.g., the Nyström method, has been applied in previous works to approx...
متن کاملLarge Scale Online Kernel Classification
In this work, we present a new framework for large scale online kernel classification, making kernel methods efficient and scalable for large-scale online learning tasks. Unlike the regular budget kernel online learning scheme that usually uses different strategies to bound the number of support vectors, our framework explores a functional approximation approach to approximating a kernel functi...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2014